24 research outputs found
Classification of Polarimetric SAR Images Using Compact Convolutional Neural Networks
Classification of polarimetric synthetic aperture radar (PolSAR) images is an
active research area with a major role in environmental applications. The
traditional Machine Learning (ML) methods proposed in this domain generally
focus on utilizing highly discriminative features to improve the classification
performance, but this task is complicated by the well-known "curse of
dimensionality" phenomena. Other approaches based on deep Convolutional Neural
Networks (CNNs) have certain limitations and drawbacks, such as high
computational complexity, an unfeasibly large training set with ground-truth
labels, and special hardware requirements. In this work, to address the
limitations of traditional ML and deep CNN based methods, a novel and
systematic classification framework is proposed for the classification of
PolSAR images, based on a compact and adaptive implementation of CNNs using a
sliding-window classification approach. The proposed approach has three
advantages. First, there is no requirement for an extensive feature extraction
process. Second, it is computationally efficient due to utilized compact
configurations. In particular, the proposed compact and adaptive CNN model is
designed to achieve the maximum classification accuracy with minimum training
and computational complexity. This is of considerable importance considering
the high costs involved in labelling in PolSAR classification. Finally, the
proposed approach can perform classification using smaller window sizes than
deep CNNs. Experimental evaluations have been performed over the most
commonly-used four benchmark PolSAR images: AIRSAR L-Band and RADARSAT-2 C-Band
data of San Francisco Bay and Flevoland areas. Accordingly, the best obtained
overall accuracies range between 92.33 - 99.39% for these benchmark study
sites
Operational Support Estimator Networks
In this work, we propose a novel approach called Operational Support
Estimator Networks (OSENs) for the support estimation task. Support Estimation
(SE) is defined as finding the locations of non-zero elements in a sparse
signal. By its very nature, the mapping between the measurement and sparse
signal is a non-linear operation. Traditional support estimators rely on
computationally expensive iterative signal recovery techniques to achieve such
non-linearity. Contrary to the convolution layers, the proposed OSEN approach
consists of operational layers that can learn such complex non-linearities
without the need for deep networks. In this way, the performance of the
non-iterative support estimation is greatly improved. Moreover, the operational
layers comprise so-called generative \textit{super neurons} with non-local
kernels. The kernel location for each neuron/feature map is optimized jointly
for the SE task during the training. We evaluate the OSENs in three different
applications: i. support estimation from Compressive Sensing (CS) measurements,
ii. representation-based classification, and iii. learning-aided CS
reconstruction where the output of OSENs is used as prior knowledge to the CS
algorithm for an enhanced reconstruction. Experimental results show that the
proposed approach achieves computational efficiency and outperforms competing
methods, especially at low measurement rates by a significant margin. The
software implementation is publicly shared at
https://github.com/meteahishali/OSEN
Convolutional Sparse Support Estimator Network (CSEN) From energy efficient support estimation to learning-aided Compressive Sensing
Support estimation (SE) of a sparse signal refers to finding the location
indices of the non-zero elements in a sparse representation. Most of the
traditional approaches dealing with SE problem are iterative algorithms based
on greedy methods or optimization techniques. Indeed, a vast majority of them
use sparse signal recovery techniques to obtain support sets instead of
directly mapping the non-zero locations from denser measurements (e.g.,
Compressively Sensed Measurements). This study proposes a novel approach for
learning such a mapping from a training set. To accomplish this objective, the
Convolutional Support Estimator Networks (CSENs), each with a compact
configuration, are designed. The proposed CSEN can be a crucial tool for the
following scenarios: (i) Real-time and low-cost support estimation can be
applied in any mobile and low-power edge device for anomaly localization,
simultaneous face recognition, etc. (ii) CSEN's output can directly be used as
"prior information" which improves the performance of sparse signal recovery
algorithms. The results over the benchmark datasets show that state-of-the-art
performance levels can be achieved by the proposed approach with a
significantly reduced computational complexity
Hyperspectral Image Analysis with Subspace Learning-based One-Class Classification
Hyperspectral image (HSI) classification is an important task in many
applications, such as environmental monitoring, medical imaging, and land
use/land cover (LULC) classification. Due to the significant amount of spectral
information from recent HSI sensors, analyzing the acquired images is
challenging using traditional Machine Learning (ML) methods. As the number of
frequency bands increases, the required number of training samples increases
exponentially to achieve a reasonable classification accuracy, also known as
the curse of dimensionality. Therefore, separate band selection or
dimensionality reduction techniques are often applied before performing any
classification task over HSI data. In this study, we investigate recently
proposed subspace learning methods for one-class classification (OCC). These
methods map high-dimensional data to a lower-dimensional feature space that is
optimized for one-class classification. In this way, there is no separate
dimensionality reduction or feature selection procedure needed in the proposed
classification framework. Moreover, one-class classifiers have the ability to
learn a data description from the category of a single class only. Considering
the imbalanced labels of the LULC classification problem and rich spectral
information (high number of dimensions), the proposed classification approach
is well-suited for HSI data. Overall, this is a pioneer study focusing on
subspace learning-based one-class classification for HSI data. We analyze the
performance of the proposed subspace learning one-class classifiers in the
proposed pipeline. Our experiments validate that the proposed approach helps
tackle the curse of dimensionality along with the imbalanced nature of HSI
data
R2C-GAN: Restore-to-Classify GANs for Blind X-Ray Restoration and COVID-19 Classification
Restoration of poor quality images with a blended set of artifacts plays a
vital role for a reliable diagnosis. Existing studies have focused on specific
restoration problems such as image deblurring, denoising, and exposure
correction where there is usually a strong assumption on the artifact type and
severity. As a pioneer study in blind X-ray restoration, we propose a joint
model for generic image restoration and classification: Restore-to-Classify
Generative Adversarial Networks (R2C-GANs). Such a jointly optimized model
keeps any disease intact after the restoration. Therefore, this will naturally
lead to a higher diagnosis performance thanks to the improved X-ray image
quality. To accomplish this crucial objective, we define the restoration task
as an Image-to-Image translation problem from poor quality having noisy,
blurry, or over/under-exposed images to high quality image domain. The proposed
R2C-GAN model is able to learn forward and inverse transforms between the two
domains using unpaired training samples. Simultaneously, the joint
classification preserves the disease label during restoration. Moreover, the
R2C-GANs are equipped with operational layers/neurons reducing the network
depth and further boosting both restoration and classification performances.
The proposed joint model is extensively evaluated over the QaTa-COV19 dataset
for Coronavirus Disease 2019 (COVID-19) classification. The proposed
restoration approach achieves over 90% F1-Score which is significantly higher
than the performance of any deep model. Moreover, in the qualitative analysis,
the restoration performance of R2C-GANs is approved by a group of medical
doctors. We share the software implementation at
https://github.com/meteahishali/R2C-GAN
Improved Active Fire Detection using Operational U-Nets
As a consequence of global warming and climate change, the risk and extent of
wildfires have been increasing in many areas worldwide. Warmer temperatures and
drier conditions can cause quickly spreading fires and make them harder to
control; therefore, early detection and accurate locating of active fires are
crucial in environmental monitoring. Using satellite imagery to monitor and
detect active fires has been critical for managing forests and public land.
Many traditional statistical-based methods and more recent deep-learning
techniques have been proposed for active fire detection. In this study, we
propose a novel approach called Operational U-Nets for the improved early
detection of active fires. The proposed approach utilizes Self-Organized
Operational Neural Network (Self-ONN) layers in a compact U-Net architecture.
The preliminary experimental results demonstrate that Operational U-Nets not
only achieve superior detection performance but can also significantly reduce
computational complexity
Dual and single polarized sar image classification using compact convolutional neural networks
Accurate land use/land cover classification of synthetic aperture radar (SAR) images plays an important role in environmental, economic, and nature related research areas and applications. When fully polarimetric SAR data is not available, single- or dual-polarization SAR data can also be used whilst posing certain difficulties. For instance, traditional Machine Learning (ML) methods generally focus on finding more discriminative features to overcome the lack of information due to single- or dual-polarimetry. Beside conventional ML approaches, studies proposing deep convolutional neural networks (CNNs) come with limitations and drawbacks such as requirements of massive amounts of data for training and special hardware for implementing complex deep networks. In this study, we propose a systematic approach based on sliding-window classification with compact and adaptive CNNs that can overcome such drawbacks whilst achieving state-of-the-art performance levels for land use/land cover classification. The proposed approach voids the need for feature extraction and selection processes entirely, and perform classification directly over SAR intensity data. Furthermore, unlike deep CNNs, the proposed approach requires neither a dedicated hardware nor a large amount of data with ground-truth labels. The proposed systematic approach is designed to achieve maximum classification accuracy on single and dual-polarized intensity data with minimum human interaction. Moreover, due to its compact configuration, the proposed approach can process such small patches which is not possible with deep learning solutions. This ability significantly improves the details in segmentation masks. An extensive set of experiments over two benchmark SAR datasets confirms the superior classification performance and efficient computational complexity of the proposed approach compared to the competing methods. - 2019 by the authors.Scopu
SRL-SOA: Self-Representation Learning with Sparse 1D-Operational Autoencoder for Hyperspectral Image Band Selection
The band selection in the hyperspectral image (HSI) data processing is an important task considering its effect on the computational complexity and accuracy. In this work, we propose a novel framework for the band selection problem: Self-Representation Learning (SRL) with Sparse 1D-Operational Autoencoder (SOA). The proposed SLR-SOA approach introduces a novel autoencoder model, SOA, that is designed to learn a representation domain where the data are sparsely represented. Moreover, the network composes of 1D-operational layers with the non-linear neuron model. Hence, the learning capability of neurons (filters) is greatly improved with shallow architectures. Using compact architectures is especially crucial in autoencoders as they tend to overfit easily because of their identity mapping objective. Overall, we show that the proposed SRL-SOA band selection approach outperforms the competing methods over two HSI data including Indian Pines and Salinas-A considering the achieved land cover classification accuracies. The software implementation of the SRL-SOA approach is shared publicly1.acceptedVersionPeer reviewe
Convolutional Sparse Support Estimator Network (CSEN) : From Energy-Efficient Support Estimation to Learning-Aided Compressive Sensing
Support estimation (SE) of a sparse signal refers to finding the location indices of the nonzero elements in a sparse representation. Most of the traditional approaches dealing with SE problems are iterative algorithms based on greedy methods or optimization techniques. Indeed, a vast majority of them use sparse signal recovery (SR) techniques to obtain support sets instead of directly mapping the nonzero locations from denser measurements (e.g., compressively sensed measurements). This study proposes a novel approach for learning such a mapping from a training set. To accomplish this objective, the convolutional sparse support estimator networks (CSENs), each with a compact configuration, are designed. The proposed CSEN can be a crucial tool for the following scenarios: 1) real-time and low-cost SE can be applied in any mobile and low-power edge device for anomaly localization, simultaneous face recognition, and so on and 2) CSEN’s output can directly be used as “prior information,” which improves the performance of sparse SR algorithms. The results over the benchmark datasets show that state-of-the-art performance levels can be achieved by the proposed approach with a significantly reduced computational complexity.publishedVersionPeer reviewe